Markus Kofler
Markus Kofler
@shivamerla this was already very useful I was now able to deploy all components the only problem I see now is after the initial installation if I re-deploy the whole...
A reboot fixed the problem. And again I come with another one: I changed now the device list strategy of the device Plugin component to "volume-mounts" instead of "envvar" by...
@shivamerla I have fixed this now by creating a modified version of this file as a config map and by mounting this in the container-toolkit daemonset at /etc/nvidia-container-runtime/config.toml. This works...
@shivamerla Yes the problem is still existing. Actually I think the plugin validation was not yet started because also because of the crashing cuda-validator pods the nvidia-operator-validator which includes the...
@shivamerla Here is the output I get when I run nvidia-smi on one of the device-plugin pods: ``` sh-4.4# nvidia-smi Thu Sep 9 19:44:09 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 460.73.01 Driver...
@shivamerla Same behaviour so it keeps crashing with the following error message in the cuda-validation init container: ``` Failed to allocate device vector A (error code CUDA driver version is...
This feature should be in the newest version available via the commit https://github.com/open-webui/open-webui/commit/217f4daef09b36d3d4cc4681e11d3ebd9984a1a5 As with the existing injectable parameters like `__request__`, there is now an `__oauth_token__` parameter which can be...