align-anything icon indicating copy to clipboard operation
align-anything copied to clipboard

[Question] Slow Speed of vLLM when evaluating MMLU

Open cby-pku opened this issue 1 year ago • 1 comments

Required prerequisites

Questions

When evaluating MMLU, the codebase supports vLLM inference, but the speed is slow (20 minutes for a single task). According to my experience, the normal speed is 20 minutes for all tasks.

cby-pku avatar Aug 05 '24 18:08 cby-pku

Thank you for your question! This is a known issue. Since the current architecture implements the BaseInference class based on deepspeed and vllm in the same Python file, importing deepspeed-related dependencies causes vllm to fail to start properly. Therefore, I set distributed_executor_backend="ray" when starting vllm. This does significantly affect efficiency. We will further modify the framework in the next version to completely decouple the two backends and fully unleash the inference speed of vllm.

Kass123777 avatar Aug 05 '24 23:08 Kass123777