mlc-llm icon indicating copy to clipboard operation
mlc-llm copied to clipboard

Multi-GPU support for larger-than-VRAM models

Open emvw7yf opened this issue 1 year ago • 1 comments

Awesome project, thanks!

Does it support sharding large models across multiple GPUs, or would this be in scope for this project in the future?

emvw7yf avatar Apr 30 '23 15:04 emvw7yf

thank you for your suggestion, yes we love to support the needs of the community and bring it in our roadmap

tqchen avatar Apr 30 '23 16:04 tqchen

llama.cpp seems to support this now fyi!

earonesty avatar Aug 28 '23 21:08 earonesty

Multi gpu now lands in mlc

tqchen avatar Oct 24 '23 13:10 tqchen