Qwen2.5-VL icon indicating copy to clipboard operation
Qwen2.5-VL copied to clipboard

Introducing the MMWorld benchmark

Open jkooy opened this issue 1 month ago • 1 comments

Dear Qwen team,

We are a big fan of your Qwen series and noticed that you have evaluated your models on several video-language benchmarks. We were wondering if you might be interested in evaluating your models on our MMWorld benchmark (https://arxiv.org/abs/2406.08407). MMWorld is designed to assess models' reasoning capabilities across various reasoning tasks and disciplines and could serve as a useful evaluation benchmark for your model development. Thank you!

jkooy avatar Jan 27 '25 20:01 jkooy