chenhuixi

Results 2 comments of chenhuixi

Perhaps the batch size is set so large that it lead to “CUDA out of memory”, but the program does not report an error. Try to make the ”train_micro_batch_size_per_gpu“ parameter...

chatglm是prefix llm,chatglm2是causal llm。 llama系列是causal llm。