OLMo icon indicating copy to clipboard operation
OLMo copied to clipboard

Finetuning config file

Open joellliu opened this issue 1 year ago • 3 comments

❓ The question

Hi, I am wondering if you can provide your config file for finetuning on the Tulu V2 dataset? It would be helpful for reproducing the finetuning results. In addition, have you tried SFT on the 1B model? How's the result? Thank you!

joellliu avatar Jun 10 '24 15:06 joellliu

Have you looked at https://github.com/allenai/open-instruct? It might have what you're looking for.

2015aroras avatar Jun 14 '24 17:06 2015aroras

Have you looked at https://github.com/allenai/open-instruct? It might have what you're looking for.

@2015aroras Thanks for your reply! It seems the open-instruct repo is mainly for DPO. I wonder if you can provide the supervised finetuning config for training on the Tuluv2 dataset. Thank you!

joellliu avatar Jun 22 '24 18:06 joellliu

@hamishivi You probably know more about this?

2015aroras avatar Jul 16 '24 18:07 2015aroras

Hi! I have the same question and am wondering if there have been any updates on this? Thanks!

ethanlshen avatar Dec 24 '24 07:12 ethanlshen

Hi, you can find the OLMo 2 SFT and DPO configs here: https://github.com/allenai/open-instruct/tree/main/configs/train_configs/olmo2

We do SFT and DPO in open-instruct (and RLVR).

hamishivi avatar Dec 24 '24 15:12 hamishivi