codellama icon indicating copy to clipboard operation
codellama copied to clipboard

Fine-tuning Code Llama on my own code

Open g12bftd opened this issue 1 year ago • 3 comments

Thank you for this amazing effort!

I would like to fine-tune code llama on my own Python code, let's call it MyPackage for now. Ultimately, I would like to ask Code llama questions of the form: "how do I do [INSERT FEASIBLE TASK NAME] using MyPackage?"

I have my package ready and documented. How do I format the dataset to fine-tune Code Llama? Are there fine-tuning scripts available in this repository?

g12bftd avatar Oct 11 '23 12:10 g12bftd

same issue.

shatealaboxiaowang avatar Oct 25 '23 07:10 shatealaboxiaowang

Any answers for this question?

david-strejc avatar Dec 22 '23 15:12 david-strejc

Hi @g12bftd, we don’t host any fine-tuning scripts in this repository. You can check out https://github.com/facebookresearch/llama-recipes which includes fine-tuning recipes for Llama 2 models and works with Code Llama as well.

Wrt dataset formatting, this is hard to answer from the get-go. However, getting LLMs to work with data not seen in training is a frequent use case so I would expect you’d find various further resources on the web. I’d probably start with in-context learning (i.e., putting relevant or abbreviated contents of your repo in your prompt) instead of fine-tuning.

jgehring avatar Dec 23 '23 10:12 jgehring