LLaVA-NeXT icon indicating copy to clipboard operation
LLaVA-NeXT copied to clipboard

Results 315 LLaVA-NeXT issues
Sort by recently updated
recently updated
newest added

Thanks for opening source the powfer model. the case show how to generate prompt for text-image and video-image. but If I want to input image-video-text at the same time, how...

I see in the code that mm_newline_position can be selected as grid, one_token, frame, and no_token, what is the exact meaning of these parameters?

How to change the code to analyze differences between videos with similar backgrounds but different foreground objects in docs/LLaVA_OneVision_Tutorials.ipynb? ![image](https://github.com/user-attachments/assets/a1b21d5e-0fc4-4635-a8c6-dc29570d6561)

I am re-evaluating the LLaVA-OneVision 0.5B on ActivityNet-QA and trying to get the value 50.5%. I get the model checkpoints using following commands: ``` warnings.filterwarnings("ignore") pretrained = "lmms-lab/llava-onevision-qwen2-0.5b-ov" model_name =...

Great work! But I can't find the interleave_demo.py in playground/demo/ as your doc instructs.

It seems that decord is causing issues with completing installation. How can I rectify this? When I try to run: pip install -e ".[train]" I get the following: Obtaining file:///Users/bill/Documents/uni/2025%20Fall%20Semester%20/Research%20499/lLaVA-NEXT_project/LLaVA-NeXT...

Are there any scripts or a link to how to train the llava-next-llama-8B model?

Hi All, I have step up everything with LLaVA-Next repo. and I want to run the pretrain code file for one vision dataset however when I am running the code...

I found this issue when working with the lmms-lab/llava-onevision-qwen2-7b-ov model and qwen2vl.(the transformers library is the latest version.) ### Code ```python import json import argparse from PIL import Image import...