text-generation-inference icon indicating copy to clipboard operation
text-generation-inference copied to clipboard

Docker container for version 2.3.0 CUDA detection broken

Open JoeGonzalez0886 opened this issue 1 year ago • 1 comments

System Info

Running this container on multiples services produces an issue with cuda gpu detection. No gpus are detected.

  • Running LLama 3.1 from HF -Tried on Runpod/Local/Novita platforms. -GPUs tested RTX 4090, A4500.

Reverting back to container tagged version :2.2.0 Fixes the issue.

Just though I would post this up just in case others are usuing 2.3.0 in production, we had a automated scaling process instantiate the new container with :latest tagged and it brought down our production systems.

Please take a look this issue team.

Thank you.

Information

  • [X] Docker
  • [ ] The CLI directly

Tasks

  • [ ] An officially supported command
  • [ ] My own modifications

Reproduction

  1. Pull latest 2.3.0 docker images
  2. Run with any LLM.
  3. Will faill to find GPU Uploading Screenshot_2024-09-20_at_2.20.05_PM.png…

Expected behavior

We would expect this version to automatically detect local GPU cuda.

JoeGonzalez0886 avatar Sep 20 '24 23:09 JoeGonzalez0886

We ran into the same issue yesterday with our docker launching scripts using latest image tag. Looks like latest is pointing to 2.3.0-rocm tag instead of 2.3.0.

Using version based tag addressed the issue

antonpolishko avatar Sep 21 '24 18:09 antonpolishko