Only one gpu is used when deploying interactive demo
Hi, VITA-1.5 looks great! When I run python -m web_demo.server --model_path demo_VITA_ckpt --ip 0.0.0.0 --port 8081 , I expect 2 gpus to be used (loading 2 models). But it appears that only one is being used? Can you help me on that?
Besides, there is another minor issues I met and I sort of fixed it by my self:
- web_ability_demo.py and server.py has the line
config_path = os.path.join(model_path, 'origin_config.json'), but there is no 'origin_config.json'. I guess one just need to rename the original config to origin_config and copy the vllm config.
Thank you for your attention. We have updated the code. Please download the latest version to experience it.
I have the same question, it appears that only one is being used. And i have updatedd the code,have you solve the issue?
@ryansunyz We need to set the environment variables before import vLLM/Pytorch. This solution should help.