ColossalAI icon indicating copy to clipboard operation
ColossalAI copied to clipboard

[FEATURE]: please add 8-bits NLP model training mode.

Open yynil opened this issue 2 years ago • 2 comments

Describe the feature

Currently FP16 support can only make it possible for training models smaller than 2B in one graphic card with 24gb ram. However the main stream useful model size is around 6~7B, I can't even load the model in FP16 mode. If I loaded in 8bits, the training will fail due to datatype mismatched.

Would you please add the 8bits model training in the near future?

yynil avatar Apr 10 '23 06:04 yynil

After resolving the licensing issue, we may be able to provide users with some pre-trained weights. You may deploy and do inference with 8 bits then. But for now, your advice is appreciated and we will consider it.

JThh avatar Apr 13 '23 04:04 JThh

Hi @yynil We have add How to train with limited resources https://github.com/hpcaitech/ColossalAI/tree/main/applications/Chat#faq For 8bits training, we are not currently making it a high priority task due to the potential for serious convergence issues. Thanks.

binmakeswell avatar Apr 17 '23 07:04 binmakeswell