aws-virtual-gpu-device-plugin
aws-virtual-gpu-device-plugin copied to clipboard
image type latest no longer supported.

as can seen from the image, nvidia no longer keeps image with tag latest, thus creating a problem while trying to run a init container as speciied here
38 initContainers:
39 - name: set-compute-mode
40 image: nvidia/cuda:latest
41 command: ['nvidia-smi', '-c', 'EXCLUSIVE_PROCESS']
42 securityContext:
43 capabilities:
44 add: ["SYS_ADMIN"]
What is the solution for this? Is it a good idea to use image based on installed cuda version in your worker node or is there some other approach ?
PS: This is more like a bug that requires fix Also my first time opening an issue, so please correct me with the tags
Thanks for reporting the issue.
to use image based on installed cuda version in your worker node
Yes. container image cuda is good enough. Feel free to file a PR
![]()
as can seen from the image, nvidia no longer keeps image with tag latest, thus creating a problem while trying to run a init container as speciied here
38 initContainers: 39 - name: set-compute-mode 40 image: nvidia/cuda:latest 41 command: ['nvidia-smi', '-c', 'EXCLUSIVE_PROCESS'] 42 securityContext: 43 capabilities: 44 add: ["SYS_ADMIN"]
What is the solution for this? Is it a good idea to use image based on installed cuda version in your worker node or is there some other approach ?
PS: This is more like a bug that requires fix Also my first time opening an issue, so please correct me with the tags
@M-A-N-I-S-H-K I have used "nvidia/cuda:11.2.2-devel-centos8" instead of "nvidia/cuda:latest" to resolve this issue