starcoder icon indicating copy to clipboard operation
starcoder copied to clipboard

Has anyone attempted to fine-tune the Starcoder model with your own code?

Open h-clickshift opened this issue 1 year ago • 3 comments

As per the title, I have attempted to fine-tune Starcoder with my own 400MB Python code. I concatenated all .py files into a single text file, similar to the content column of the bigcode/the-stack-dedup Parquet. I then scanned the text and sliced code snippets with 1024 characters to train the model for 1000 steps. However, during validation, I could not achieve any recall of my own code.

I have been informed that fine-tuning a language model does not require a large dataset.

h-clickshift avatar May 29 '23 02:05 h-clickshift

can you test my fork

https://github.com/MHketbi/starcoder

MubarakHAlketbi avatar May 29 '23 07:05 MubarakHAlketbi

May I know the manner you fine-tune your model and the way you estimate your performance of it, i.e. with fill-in-middle task or next-code-sentence prediction task? AFA I can know, the best fine-tuning performance can be obtained when your fine-tune following their instructions on Mega repository or fine-tune with fill-in-middle task given in GPT2.

However, if you are only validating your model in a model validation step and you find you code could not be reproduced, I suspect there could be implementation error with your code. You are welcome to provide your implementation snippet here so we could check.

WrViajero avatar Jun 14 '23 08:06 WrViajero

any update on this please?

thanhnew2001 avatar Jul 23 '23 05:07 thanhnew2001