unilm
unilm copied to clipboard
【kosmos-2】The code for GRIT construction
Describe Model I am using kosmos-2: Will you updata the code of GRIT construction process? I'd like to finetune kosmos-2 in App UI scene, but the detail of GRIT construction is not clear enough for me. For example, the steps of "get noun chunks and region from detector" and "input image and noun chunks into glip to obtain bboxes" seems same? Thaks for your great work!
Oh,sorry! I made a mistake. Another question: The process of generate grit is strict? In a specific scene, GLIP might can not recognize all objects. In this case, is it possible to generate object bbox、captions and nuon-chunks manually for fientuning?
Yes, manual annotations would be quite helpful.
Hi! I am also curious about the construction of the GRIT dataset. It is mentioned in the paper that
We eliminate certain abstract noun phrases that are challenging to recognize in the image, such as “time”, “love”, and “freedom”, to reduce potential noise.
So, the abstract noun phrases are eliminated manually or using spacy? Many thanks!