LAVIS
LAVIS copied to clipboard
How to use your own dataset to train and fine-tune the VQA task of BLIP2-flant5xl
Hi, thank you very much for open source. I want to use my own Image and caption, and QA data to fine-tune the BLIP2 data. Should my process be to prepare the same data set for okvaq, and then run the /run_scripts/blip2/eval/eval_okvqa_zeroshot_flant5xl.sh file? Then should I copy evaluate.py into the run_scripts/blip2/eval/ path? Or is my approach wrong?
Hi, eval_okvqa_zeroshot_flant5xl.sh provides the script for evaluation. You can refer to train_caption_coco.sh for fine-tuning on image captioning. We are still working on providing support for VQA fine-tuning.
Thanks.
Thank you very much for your comments. However, it appears that the codes are designed for fine-tuning on the COCO dataset rather than a custom dataset. I was wondering if it would be possible to make modifications to the code in order to fine-tune the model on our custom dataset by registering it in the 'builders' directory?
@chenyd0763, can you take a look at our tutorial on how to add new datasets? https://opensource.salesforce.com/LAVIS//latest/tutorial.datasets.html
Hi, eval_okvqa_zeroshot_flant5xl.sh provides the script for evaluation. You can refer to train_caption_coco.sh for fine-tuning on image captioning. We are still working on providing support for VQA fine-tuning.
Thanks.
looking forward to the training and finetuning code
Thank to your response, I will try it later.
魇 @.***
------------------ 原始邮件 ------------------ 发件人: "salesforce/LAVIS" @.>; 发送时间: 2023年3月14日(星期二) 下午2:16 @.>; @.@.>; 主题: Re: [salesforce/LAVIS] How to use your own dataset to train and fine-tune the VQA task of BLIP2-flant5xl (Issue #152)
Hi, eval_okvqa_zeroshot_flant5xl.sh provides the script for evaluation. You can refer to train_caption_coco.sh for fine-tuning on image captioning. We are still working on providing support for VQA fine-tuning.
Thanks.
looking forward to the training and finetuning code
— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you authored the thread.Message ID: @.***>
Also looking forward to the training and fine-tuning code.
Is fine-tuning code available?
looking forward to the training and finetuning code
looking forward to the training and finetuning code
Looking forward to the finetuning code for VQA, think it could lead to some very interesting applications :)
looking forward to the fine-tuning code for VQA as well.
looking forward to the fine-tuning code for VQA +1
Also looking forward to the fine-tuning support. Is it here yet? :)
Also looking forward to the fine-tuning support!
Also looking forward to the fine-tuning code on VQA!
VQA的finetune还没出来吗
Also looking forward to the fine-tuning code for VQA!
Looking forward to fine-tuning for VQA!
looking forward to fine tuning for vqa. at this point just captioning and running llm of choice, but obviously will be awesome if vqa can be fine tuned directly
Also looking forward to the fine-tuning code for VQA :)
does anybody know if code for BLIP2 VQA finetuning is available? /thanks
I know, no, obviously.
Hi everyone. I have implemented the BLIP-VQA-BASE model for the VQA task here. I hope this implementation can help you and this implementation will receive some advice.
does anybody know if code for BLIP2 VQA finetuning is available? /thanks
hi, have you implemented fine-tune blip2 on the vqa task?