Open-LLaVA-NeXT icon indicating copy to clipboard operation
Open-LLaVA-NeXT copied to clipboard

An open-source implementation for training LLaVA-NeXT.

Results 8 Open-LLaVA-NeXT issues
Sort by recently updated
recently updated
newest added

I wanted to know why the `prepare_input_labels_for_multimodal` function in `llava_arch.py` is designed to [throw an exception ](https://github.com/xiaoachen98/Open-LLaVA-NeXT/blob/master/llava/model/llava_arch.py#L205)during pretraining if the `mm_use_im_start_end` option is enabled: ```python # TODO: image start /...

Thanks for your great work! I'm wondering if u can share the loss curve for training llava-next-llama3? I've observed some different behaviour compared to training llava-next-vicuna-7b. I'm wondering if it's...

Hi, I got a training curve like this, is it normal? Do you mind sharing your trainer_state.json? thx!

What is the difference between the anyres implementation in this project and the S2-based implementation in llava?

I conducted experiments on lllama3-8b. This repository can achieve comparable performance to lllama-next-llama3-8B on MME (1613 vs. 1603), but not on MMMU (37.4 vs. 41.7). Do you have any ideas?

I found that the 'self.image_newline' in LlavaMetaModel was not tuned during the pre-training stage. Have you tried enabling this parameter during pre-training?

Hi, thank you for sharing this repo. I've been working on preparing the SFT dataset, but I've encountered an issue with collecting all the necessary data. It seems that the...