transformer-deploy icon indicating copy to clipboard operation
transformer-deploy copied to clipboard

convert_model command not found

Open pint1022 opened this issue 2 years ago • 3 comments

Hello get docker image 0.6.0. Just tried to run the two demo command:

  1. docker run -it --rm --gpus all
    -v $PWD:/project ghcr.io/els-rd/transformer-deploy:0.6.0
    bash -c "cd /project &&
    convert_model -m "philschmid/MiniLM-L6-H384-uncased-sst2"
    --backend tensorrt onnx
    --seq-len 16 128 128" got the error: convert_model not found

  2. tried the 2nd command to use triton inference server. the service is started fine; the query command got the error: curl -X POST http://localhost:8000/v2/models/transformer_onnx_inference/versions/1/infer
    --data-binary "@demo/infinity/query_body.bin"
    --header "Inference-Header-Content-Length: 161" {"error":"Request for unknown model: 'transformer_tensorrt_inference' is not found"}

Does infinity download huggingface models and convert to Triton format?

thanks

pint1022 avatar May 23 '23 18:05 pint1022

I also got the convert_model not found.

I think pip3 install ".[GPU]" ... was lost somewhere in the latest Dockerfile update: link

I was able to run convert_model after manually running pip install ".[GPU]" inside the container

sc0eur avatar Jul 31 '23 13:07 sc0eur

I ran into the same error and fixed it as suggested. That is, manually run pip install ".[GPU]" inside the container.

jingzhaoou avatar Mar 12 '24 23:03 jingzhaoou

This issue is still there? Can't someone fix it? EDIT: it seems this repository is not being maintained. It is a shame!

aidanrussell avatar May 02 '24 15:05 aidanrussell