starcoder icon indicating copy to clipboard operation
starcoder copied to clipboard

Hardware requirements for inference and fine tuning.

Open Vipitis opened this issue 1 year ago • 1 comments

Hey,

I am finishing a project on evaluating code language models on "creative" programming (shadercode). Just yesterday I finished fine-tuning sanatacoder on three different datasets to evaluate on my metric.

Now this new project popped up but it's vastly larger. Prohibitively so. I was unable to run 6B models on the RTX A5000 I have access to. For finetuning santacoder (no_fp16, batch_size 2 and sequence length of 2048) 97% of the 24GB VRAM was used using a slightly adapted version of the provided script.

Do you have any numbers on what requirements there are for PEFT on this model? And is there a set way to run inference with the hardware I listed?

Thanks for releasing yet another interesting model!

Vipitis avatar May 07 '23 15:05 Vipitis

found #6 as closed which gives some information about inference.

Perhaps add that information to the readme

Vipitis avatar May 07 '23 17:05 Vipitis

Thanks for the feedback, this PR adds them

loubnabnl avatar May 25 '23 16:05 loubnabnl

Thanks for the hardware requirement info! It seems only have the requirement for inference. I wonder if there is anything for fine-tuning on downsteam tasks?

ywen666 avatar May 28 '23 06:05 ywen666