zyandtom
zyandtom
@kaikaishen hi have you convert this model to c++ sucessfully? actually I'm doing similar thing, I try to export the pt file in torch.
hi, I got the same question. Have you solved this problem?
`tie_word_embeddings` set true in confg.json will solve this
请问解决了吗,same error,120w video training data
可以试试换一个video读取库,decorder
also, i found InternVL3 has integrated V2PE, is there any change for visual preprocess?
> [@zyandtom](https://github.com/zyandtom) Hi! Could you find where the V2PE is implemented in InternVL3? I couldn't locate it :( i didn't find it in code too, i just saw it in...
> Yes, InternVL3 is compatible with the training and inference code of InternVL2.5 — you can directly use the same codes without modification. Currently, V2PE is not integrated into the...
hi, May I ask where this grounding dataset comes from? Also, does llava onevision have any special tokens for bboxes, or is there any preprocessing func for box? I didn't...
i have this question too. i think `lmms-lab/qwen2-0.5b-ov` may only have llm weights, because when i tried to load vision tower from `lmms-lab/qwen2-0.5b-ov`, it failed. But paper saying, at one-vision...