all-seeing
all-seeing copied to clipboard
[ICLR 2024 & ECCV 2024] The All-Seeing Projects: Towards Panoptic Visual Recognition&Understanding and General Relation Comprehension of the Open World"
I run into missing modules (llava) when I run your provided scripts since llava module is not installed. I tried running pip install . -e in the /llava directory but...
二阶段微调错误
我按照文档准备finetune第二阶段,直接运行sh scripts_asmv2/stage2-finetune.sh,报了下边的错: ValueError: Looks like distributed multinode run but MASTER_ADDR env not set, please try exporting rank 0's hostname as MASTER_ADDR 然后我把命令改成torchrun --master_port=xxxxx,结果报了CUDA Out of memory的错(即使我已经把bacthsize设成1了),环境是A100+deepspeed zero2,请问这是怎么回事
I ran sh scripts_asmv2/eval/psg_eval.sh OpenGVLab/ASMv2. I get 2 errors: **RuntimeError: Internal: could not parse ModelProto from OpenGVLab/ASMv2/tokenizer.model** and **ZeroDivisionError: division by zero** **I am not using Docker.** (/s/red/a/nobackup/vision/anju/allseeing/cvenv) carnap:/s/red/a/nobackup/vision/anju/allseeing/all-seeing-main/all-seeing-v2$ sh...
To generate the dataset with predicates, which are represented as predicate, we need to parse the predicate from long sentences (i.e., captions). How did you parse these? did human manually...
Where can the file "asmv2-13b.jsonl" be downloaded? Sincerely need help! @shepnerd @whai362 @orashi @czczup @JustinYuu
I really want to try your model, can you provide demo or any inference code guideline for us to use your code?
Hello, thank you very much for your excellent work. I would like to use your model for some image captioning tasks. Could you please provide some usage instructions? Thank you!
Thank you for the excellent work on ASMv2. In the paper, you mention that when creating the AS-V2 dataset, the bounding boxes of objects are used as part of the...
The GPU memory consumption of the model was too high, so I converted it to a LLAMA.CPP file. The GPU memory usage is fine. However, due to the nature of...