k8s.io
k8s.io copied to clipboard
Deprecate and migrate away from gs://kubernetes-release
Part of umbrella issue to migrate the kubernetes project away from use of GCP project google-containers: https://github.com/kubernetes/k8s.io/issues/1571
This issue covers the deprecation of and migration away from the following google.com assets:
- the google.com-owned GCS bucket
gs://kubernetes-release
living in GCP projectgoogle-containers
, in favor of the community-owned GCS bucketgs://k8s-release
living in GCP project TBD (currentlyk8s-release
) - the region-specific GCS buckets
gs://kubernetes-release-asia
andgs://kubernetes-release-eu
, same as above butgs://k8s-release-eu
andgs://k8s-release-asia
instead - TODO: are there container images involved here as well, or did we already address that with k8s.gcr.io?
These are not labeled as steps just yet because not everything needs to be completed to full fidelity in strict sequential order. I would prefer that we get a sense sooner rather than later what the impact of shifting dl.k8s.io traffic will be, in terms of how much budget, and what percentage of traffic that represents vs. hardcoded traffic.
Determine new-to-deprecated sync implementation and deprecation window
There are likely a lot of people out there that have gs://kubernetes-release
hardcoded. It's unreasonable to stop putting new releases there without some kind of advance warning. So after announcing our intent to deprecate gs://kubernetes-release
, we should decide how we're going to sync new releases back there (and its region-specific buckets)
-
gsutil rsync
- Google Cloud Storage Transfer Service
- etc.
As for the deprecation window itself, I think it's fair to treat this with a deprecation clock equivalent to disabling a v1 API.
Determine gs://k8s-release project location and geo-sync implementation
- Someone (probably me) manually created
gs://k8s-release
and its other buckets to prevent someone else from grabbing the name - The
-eu
and-asia
buckets are not actually region-specific, and should be recreated as such - We should decide how we're going to implement region syncing (same as above)
- We should decide at this stage whether we want to block on a binary artifact promotion process, or get by with one of the syncing mechanisms from above
Use dl.k8s.io where possible and identify remaining hardcoded bucket name references across the project
The only time a kubernetes release artifact GCS bucket name needs to show up in a URI is if gsutil is involved, or someone is explicitly interested in browsing the bucket. For tools like curl
or wget
that retrieve binaries via HTTP, we have https://dl.k8s.io
, which will allow us to automatically shift traffic from one bucket to the next depending on the requested URIs
I started doing this for a few projects while working on https://github.com/kubernetes/k8s.io/issues/2318, e.g.
- https://github.com/kubernetes/cloud-provider-gcp/pull/252
- https://github.com/kubernetes-sigs/cluster-api/pull/4958
TODO: a cs.k8s.io query and resulting checklist of repos to investigate
Shift dl.k8s.io traffic to gs://k8s-release-dev
TODO: there is a separate issue for this.
We will pre-seed gs://k8s-release with everything in gs://kubernetes-release, and gradually modify dl.k8s.io to redirect more and more traffic to gs://k8s-release.
The idea is not to flip a switch, just in case that sends us way more traffic than our budget is prepared to handle. Instead, let's consider shifting traffic gradually for certain URI patterns, or a certain percentage of requests, etc. It's unclear whether this will be as straightforward as adding lines to nginx, or whether we'll want GCLB changes as well.
Change remaining project references to gs://k8s-release
/area artifacts /area prow /area release-eng /sig release /sig testing /wg k8s-infra /priority important-soon /kind cleanup /milestone v1.23
/cc @kubernetes/release-engineering
Blocked on https://github.com/kubernetes/k8s.io/issues/1375
/milestone v1.24
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle stale
- Mark this issue or PR as rotten with
/lifecycle rotten
- Close this issue or PR with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
/remove-lifecycle stale
/milestone clear /lifecycle frozen /priority backlog
/remove-lifecycle frozen /milestone v1.26 /priority important-longterm
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle stale
- Mark this issue or PR as rotten with
/lifecycle rotten
- Close this issue or PR with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
/remove-lifecycle stale /milestone v1.27
Blocked by https://github.com/kubernetes/k8s.io/issues/4528
/milestone v1.28 /lifecycle frozen
This isn't blocked by 4528, everything can switch to only use dl.k8s.io immediately.
Lots of hits still https://cs.k8s.io/?q=%2Fkubernetes-release&i=nope&files=&excludeFiles=&repos=
/assign
Below are all the references for https://storage.googleapis.com/kubernetes-release
that need to be updated. This list wont include the changes that will be necessary for gs://kubernetes-release
updates. I'll generate and track those changes next.
- [x] kubernetes/kops | search results | https://github.com/kubernetes/kops/issues/15402
- [x] kubernetes/test-infra | search results | https://github.com/kubernetes/test-infra/pull/29495
- [x] kubernetes/website | search results | https://github.com/kubernetes/website/issues/41069
- [x] kubernetes/kubernetes | search results | https://github.com/kubernetes/kubernetes/issues/117949 CHANGELOG will not be updated
- [x] kubernetes/cloud-provider-gcp | search results | https://github.com/kubernetes/cloud-provider-gcp/pull/540
- [x] kubernetes-sigs/kubespray | search results | https://github.com/kubernetes-sigs/kubespray/pull/10066
- [x] kubernetes-csi/csi-driver-smb | search results | https://github.com/kubernetes-csi/csi-driver-smb/pull/614
- [x] kubernetes-sigs/kui | search results | https://github.com/kubernetes-sigs/kui/pull/9360
- [x] kubernetes-sigs/aws-ebs-csi-driver | search results | https://github.com/kubernetes-sigs/aws-ebs-csi-driver/pull/1593
- [x] kubernetes-sigs/cluster-api-provider-digitalocean | search results | https://github.com/kubernetes-sigs/cluster-api-provider-digitalocean/pull/490
- [x] kubernetes-sigs/kind | search results | https://github.com/kubernetes-sigs/kind/pull/3217
- [x] kubernetes-sigs/krew | search results | https://github.com/kubernetes-sigs/krew/pull/824
- [x] kubernetes-sigs/secrets-store-csi-driver | search results | https://github.com/kubernetes-sigs/secrets-store-csi-driver/pull/1255
- [x] kubernetes/ingress-nginx | search results | https://github.com/kubernetes/ingress-nginx/pull/9946
- [x] ~~kubernetes/k8s.io | search results~~ This repo does not need to be updated
- [x] kubernetes/kubeadm | search results | https://github.com/kubernetes/kubeadm/pull/2872
- [x] kubernetes-client/haskell | search results | https://github.com/kubernetes-client/haskell/pull/103
- [x] kubernetes-client/java | search results | https://github.com/kubernetes-client/java/pull/2657
- [x] kubernetes-client/python | search results | https://github.com/kubernetes-client/python/pull/2055
- [x] kubernetes-csi/csi-driver-nfs | search results | https://github.com/kubernetes-csi/csi-driver-nfs/pull/450
- [x] kubernetes-sigs/apiserver-network-proxy | search results | https://github.com/kubernetes-sigs/apiserver-network-proxy/pull/492
- [x] kubernetes-sigs/azuredisk-csi-driver | search results | https://github.com/kubernetes-sigs/azuredisk-csi-driver/pull/1845
- [x] kubernetes-sigs/azurefile-csi-driver | search results | https://github.com/kubernetes-sigs/azurefile-csi-driver/pull/1250
- [x] kubernetes-sigs/azurelustre-csi-driver | search results | https://github.com/kubernetes-sigs/azurelustre-csi-driver/pull/129
- [x] kubernetes-sigs/blob-csi-driver | search results | https://github.com/kubernetes-sigs/blob-csi-driver/pull/932
- [x] kubernetes-sigs/cli-experimental | search results | https://github.com/kubernetes-sigs/cli-experimental/pull/340
- [x] kubernetes-sigs/cloud-provider-azure | search results | https://github.com/kubernetes-sigs/cloud-provider-azure/pull/3878
- [x] kubernetes-sigs/cluster-api-provider-aws | search results | https://github.com/kubernetes-sigs/cluster-api-provider-aws/pull/4263
- [x] kubernetes-sigs/cluster-api-provider-azure | search results | https://github.com/kubernetes-sigs/cluster-api-provider-azure/pull/3559
- [x] kubernetes-sigs/cluster-api-provider-nested | search results | https://github.com/kubernetes-sigs/cluster-api-provider-nested/pull/346
- [x] kubernetes-sigs/cluster-api-provider-openstack | search results | https://github.com/kubernetes-sigs/cluster-api-provider-openstack/pull/1542
- [x] kubernetes-sigs/cluster-api-provider-vsphere | search results | https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/pull/1910
- [x] kubernetes-sigs/descheduler | search results | https://github.com/kubernetes-sigs/descheduler/pull/1145
- [x] kubernetes-sigs/krew-index | search results | https://github.com/kubernetes-sigs/krew-index/pull/3128
- [x] kubernetes-sigs/nfs-ganesha-server-and-external-provisioner | search results | https://github.com/kubernetes-sigs/nfs-ganesha-server-and-external-provisioner/pull/130
- [x] kubernetes-sigs/prometheus-adapter | search results | https://github.com/kubernetes-sigs/prometheus-adapter/pull/583
- [x] kubernetes-sigs/windows-gmsa | search results | https://github.com/kubernetes-sigs/windows-gmsa/pull/108
- [x] kubernetes/cloud-provider-openstack | search results | https://github.com/kubernetes/cloud-provider-openstack/pull/2240
- [x] kubernetes/cloud-provider-vsphere | search results | https://github.com/kubernetes/cloud-provider-vsphere/pull/722
- [x] kubernetes/dashboard | search results | https://github.com/kubernetes/dashboard/pull/7729
- [x] ~~kubernetes/design-proposals-archive | search results~~ Project is Archived
- [x] kubernetes/kube-state-metrics | search results | https://github.com/kubernetes/kube-state-metrics/pull/2063
- [x] kubernetes/minikube | search results | https://github.com/kubernetes/minikube/pull/16487
- [x] kubernetes/sig-release | search results | https://github.com/kubernetes/sig-release/pull/2241
Here are the results for gs://kubernetes-release
. These changes will need to be a bit more involved and careful as:
if those references are reading files they should switch to using https / curl / wget if they're writing files (kubernetes release) we can't migrate yet
- [x] kubernetes/kubernetes | search results | https://github.com/kubernetes/kubernetes/pull/118082 NOTE: One ref in cluster/gce/gci/mounter/stage-upload.sh is used to write content
- [x] kubernetes/cloud-provider-gcp | search results | https://github.com/kubernetes/cloud-provider-gcp/pull/540
- [x] kubernetes/release | search results | Waiting on https://github.com/kubernetes/release/issues/3065
- [x] kubernetes/sig-release | search results | https://github.com/kubernetes/sig-release/pull/2243
- [x] kubernetes-sigs/kubetest2 | search results |https://github.com/kubernetes-sigs/kubetest2/pull/228
- [x] kubernetes-sigs/tejolote | search results | Tejolote uses GCS to publish artifacts, I don't believ any updates will be needed to support this effort since they write content at this time, a ticket was opened to track supporting https uploads for the future https://github.com/kubernetes-sigs/tejolote/issues/100
- [x] kubernetes/community | search results | https://github.com/kubernetes/community/pull/7301
- [x] kubernetes/kubeadm | search results | https://github.com/kubernetes/kubeadm/pull/2872
What is the plan for gs://k8s-release-dev/ci/
? The plumbing for https://dl.k8s.io/ci
seem to be in place already. Are folks good to start moving to that ref?
k8s-release-dev will be tracked separately and isn't meant to be end-user facing (unlike dl.k8s.io), it's meant for contributors to the project
kubernetes/release | search results Cannot be updated ATT, changes are used to push artifacts to the bucket
One of them is NOT:
https://github.com/kubernetes/release/blob/065c82ea4a3ca8f0e4b1b87ade902cb9e18be78d/hack/rapture/publish-packages.sh#L60
This is consuming release binaries.
Even most of the release tools should be updated unless they're writing content, and only when writing. We'll deal with writing content later.
I believe at this point all remaining references either have justification for not being updated or are actively awaiting PRs to be reviewed.
Awesome, thank you! I'll plan to take a pass through remaining references again when the PRs are in.
@BenTheElder, I think we're ready for a check on outstanding references.
Looking at:
the remaining references will go away (except the blog posts and changelogs) after the redirect update and the migration to a community-owned release bucket.
/milestone v1.29
Though, for the moment we're still seeing most bandwidth / requests go to the bucket.
Hopefully when we start publishing only to a new kubernetes.io bucket we'll see that start to change.
/milestone v1.30
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue as fresh with
/remove-lifecycle stale
- Close this issue with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
/remove-lifecycle stale
/milestone v1.31
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue as fresh with
/remove-lifecycle stale
- Close this issue with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale