zyandtom

Results 19 comments of zyandtom

@kaikaishen hi have you convert this model to c++ sucessfully? actually I'm doing similar thing, I try to export the pt file in torch.

hi, I got the same question. Have you solved this problem?

`tie_word_embeddings` set true in confg.json will solve this

请问解决了吗,same error,120w video training data

可以试试换一个video读取库,decorder

also, i found InternVL3 has integrated V2PE, is there any change for visual preprocess?

> [@zyandtom](https://github.com/zyandtom) Hi! Could you find where the V2PE is implemented in InternVL3? I couldn't locate it :( i didn't find it in code too, i just saw it in...

> Yes, InternVL3 is compatible with the training and inference code of InternVL2.5 — you can directly use the same codes without modification. Currently, V2PE is not integrated into the...

hi, May I ask where this grounding dataset comes from? Also, does llava onevision have any special tokens for bboxes, or is there any preprocessing func for box? I didn't...

i have this question too. i think `lmms-lab/qwen2-0.5b-ov` may only have llm weights, because when i tried to load vision tower from `lmms-lab/qwen2-0.5b-ov`, it failed. But paper saying, at one-vision...