Multi-Modality-Arena icon indicating copy to clipboard operation
Multi-Modality-Arena copied to clipboard

Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP...

Results 20 Multi-Modality-Arena issues
Sort by recently updated
recently updated
newest added

Hi, I'm the author of the paper [Aligning Large Multi-Modal Model with Robust Instruction Tuning](https://github.com/FuxiaoLiu/LRV-Instruction) and want to add our model to your amazing arena. May I know you email...

Nice work! Interested in the design of 1 vs 1 battles between LVLMs, but can you share more details about the Elo rating algorithm? Like the choice of k-factor, the...

Hi! I'm a fan of your work. Can you please provide more details about how to do eval for MiniGPT-4 and LLaVA on various datasets? Thanks a lot!

Thanks for your work! After reading the paper OmniMedVQA, I have two questions and sincerely look forward to the answers. 1. From the paper of MedVInT and RadFM, the dataset...

Can you share the application link for these datasets?

Check http://lvlm-ehub.opengvlab.com/leaderboard.html but found:

Could you please provide the link for QA_PMC_LLaMA_lora_PMC-CLIP_MLP/choice_training/checkpoint-4146 you use in this project (Multi-Modality-Arena/MedicalEval/Question-answering_Score/MedVInT/src/MedVInT_TD/test.py)?

the link of AVN Assessment dataset in your paper OmnimedVQA has expired. can you please update the url? thank you Thivya Narendran. Image set for retinal arteryvein nicking assessment. https://people.eng.unimelb.edu.au/thivun/projects/AV_nicking_quantification/...

How can I download datasets in OmnimedVQA? I want to us the 3D Modality dataset in your paper bust there is no links