LAVIS
LAVIS copied to clipboard
Fine-tuning InstructBLIP?
It seems that their fine-tuning strategy is similar to the standard training approach for VQA. I noticed that the blip2_vicuna_instruct.py
file includes a predict_answers
function, which is commonly used in VQA tasks.
To use their approach, you can prepare your datasets as they've described, including image
, text_input
, and text_output
, and then launch train.py
. However, I would also like to see more training details to better understand their methodology.
It seems the run_scripts do not include anything related to InstuctBlip. Will the official release the code for pre-training and fine-tuning of InstructBlip?
Bump, could you expand on the Instruction-Tuning section for the InstructBLIP model page? It's not clear how to do this or even if the full code necessary has been released.
Could you release VQA finetuning script for instructBLIP?
Could you release VQA finetuning script for instructBLIP?
meet the same problem
same question
same question
same question,how to fine tune instruct blip?
I would also like to know how to fine tune instruct blip
Could the author provide the fine-tuning script of InstructBLIP? @LiJunnan1992
mark!
mark
mark
mark!
same question here
mark
same question
+1
+1, pretrain vicuna at stage2 also need to be modified
mark
+1
mark
mark
Guys, are there any training scripts for InstructBLIP on captioning (not VQA) tasks? Something like https://github.com/salesforce/LAVIS/blob/main/run_scripts/blip2/train/train_caption_coco.sh
but for InstructBLIP?
mark
mark
mark!
https://github.com/AttentionX/InstructBLIP_PEFT?tab=readme-ov-file It semms like the authors have no motivatioon to release the finetuning script, but here is a repo that claims to do the same
We have released the finetuning scripts, so let me know if you have any problem!
We have released the finetuning scripts, so let me know if you have any problem!
Could you please tell me where it is? I found no finetuning scripts for instructblip under run_scripts/train/BLIP2