xtuner
xtuner copied to clipboard
Does xtuner support DPO for InternVL?
I am trying to do a custom DPO fine-tuning for internvl_v2_internlm2_2b_lora_finetune, but the default config is oriented towards vanilla supervised fine-tuning with images. I tried to compare / incorporate changes from internlm2_chat_1_8b_dpo_full but am running into some issues with the dataset formats supported.
Is this something that xtuner actually supports at the moment?
https://github.com/hhaAndroid/xtuner/blob/hha_0919/my_llava/README.md#dpo