Chintan
Chintan
I am facing the same issue. Has anyone found the fix?
> I am facing the same issue. Has anyone found the fix? Following https://github.com/pytorch/fairseq/issues/292 fixed the problem
> If you are still having the problem please use this code instead of above in place of tranform. > transform = transforms.Compose([transforms.ToTensor(), > transforms.Normalize((0.5,), (0.5,)) > ]) Can you...
I am facing the same issue!
I am not sure how to get the exact image version, please help me with it.
I used the `:latest` tag and all is sorted now.
``` model= #path to model or hugging face path volume=$PWD docker run --gpus all --shm-size 24g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:latest --model-id $model --max-total-tokens 5024 --max-input-length 4096 --num-shard 4 --max-concurrent-requests...
+1 Facing the same issue.
You can also add HF_HUB_ENABLE_HF_TRANSFER=0 in the docker command, `docker run --shm-size 1g --env HF_HUB_ENABLE_HF_TRANSFER=0 ....... `
With 1.1.0 it is not working, ``` model=sigmareaver/flan-ul2-4bit-128g-gptq volume=$PWD/flan-ul2-4bit-128g-gptq-data docker run --gpus all --shm-size 24g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.1.0 --model-id $model --max-total-tokens 5024 --max-input-length 4096 --num-shard 4 --max-concurrent-requests 128...