Chintan

Results 13 comments of Chintan

I am facing the same issue. Has anyone found the fix?

> I am facing the same issue. Has anyone found the fix? Following https://github.com/pytorch/fairseq/issues/292 fixed the problem

> If you are still having the problem please use this code instead of above in place of tranform. > transform = transforms.Compose([transforms.ToTensor(), > transforms.Normalize((0.5,), (0.5,)) > ]) Can you...

I am not sure how to get the exact image version, please help me with it.

``` model= #path to model or hugging face path volume=$PWD docker run --gpus all --shm-size 24g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:latest --model-id $model --max-total-tokens 5024 --max-input-length 4096 --num-shard 4 --max-concurrent-requests...

You can also add HF_HUB_ENABLE_HF_TRANSFER=0 in the docker command, `docker run --shm-size 1g --env HF_HUB_ENABLE_HF_TRANSFER=0 ....... `

With 1.1.0 it is not working, ``` model=sigmareaver/flan-ul2-4bit-128g-gptq volume=$PWD/flan-ul2-4bit-128g-gptq-data docker run --gpus all --shm-size 24g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.1.0 --model-id $model --max-total-tokens 5024 --max-input-length 4096 --num-shard 4 --max-concurrent-requests 128...