community.okd
community.okd copied to clipboard
OpenShif: Support Cluster Operator Info Gathering
Good day,
I opened an issue with the kubernetes.core which requested to add a feature that will process cluster operators in order to determine the health of a cluster.
They suggested the possibility of introducing that feature here. However, it would be almost a reimplementation of the k8s_info I think. I have a POC of the kubernetes.core in my fork without any integration tests.
I am willing to take on the work to do the PR, but I will need some guidance as to how best to implement that here.
Thanks
We had a similar issue with the raw k8s module, our solution was to put some of the meatier functionality into functions in kubernetes.k8s and import/use them in openshift.k8s (https://github.com/openshift/community.okd/blob/main/plugins/module_utils/k8s.py#L16). Unfortunately there's still plenty of duplication. You could definitely just pull in the file, or you could make a more specific clusteroperator module. With the limited logic in the k8s_info module it may make more sense to just copy it rather than trying to find a cleaner way around that.
I want to revisit this.
The k8s module has been working through a general Cluster Operator healthy statement but it doesn't help you know what the status of the individual operators are. In the event that the wait is exceeded there is still a need to know which ones are failing.
I am happy to work through this, but it goes back to, do we re-implement the k8s_info() here?
@fabianvf still interested in this
Issues go stale after 90d of inactivity.
Mark the issue as fresh by commenting /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.
Exclude this issue from closing by commenting /lifecycle frozen.
If this issue is safe to close now please do so with /close.
/lifecycle stale
Stale issues rot after 30d of inactivity.
Mark the issue as fresh by commenting /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.
Exclude this issue from closing by commenting /lifecycle frozen.
If this issue is safe to close now please do so with /close.
/lifecycle rotten /remove-lifecycle stale
I'm still interested in pushing this forward
Rotten issues close after 30d of inactivity.
Reopen the issue by commenting /reopen.
Mark the issue as fresh by commenting /remove-lifecycle rotten.
Exclude this issue from closing again by commenting /lifecycle frozen.
/close
@openshift-bot: Closing this issue.
In response to this:
Rotten issues close after 30d of inactivity.
Reopen the issue by commenting
/reopen. Mark the issue as fresh by commenting/remove-lifecycle rotten. Exclude this issue from closing again by commenting/lifecycle frozen./close
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.