mohittalele
mohittalele
@jbottum I have not tested and reproduced bug with earlier k8s version. Actually in our system we can't downgrade k8s version due to compatibility with other components such as istio.
Hi Reeve, thanks for info. If it's not that complicated, you can leave some instructions on how to fix it. Maybe I can create draft pr. On Mon, 4 Apr...
Yes there seems to be some changes in imports used in video intelligence. I will have a look and try to fix it
@shivamerla I redeployed the chart. The node was successfully scaled up. So thats a progress. However still there are no driver and container toolkit pods. Also the test pod fails...
@shivamerla I checked the Daemonsets. Its not present there. ` kubectl get daemonset -n gpu-operator` returns - ``` NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE gpu-operator-helm-node-feature-discovery-worker 2 2...
Here are the nfd-master [logs](https://gist.github.com/mohittalele/f357965726e57df88ec02a5c97957148). And nfd-daemonset did not have the toleration I specified in values.yaml. So, there was no nfd-worker pod running on the gpu node. Helm template also...
Ahh I see. That should solve the issue. I would test it next week and report back. Thanks for the prompt responses!
@shivamerla it works as expected upon adding toleration for nfd worker. We can close this issue. :)
any update ?