vision-and-language topic
LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
wikiHow_paper_list
A paper list of research conducted based on wikiHow
awesome-visual-grounding
awesome visual grounding: a curated list of research papers in referring visual grounding
VLDeformer
Pytorch implement of the paper "VLDeformer: Vision Language Decomposed Transformer for Fast Cross-modal Retrieval", KBS 2022
zeroshot-storytelling
Github repository for Zero Shot Visual Storytelling
NvEM
[ACM MM 2021 Oral] Official repo of "Neighbor-view Enhanced Model for Vision and Language Navigation"
visual-spatial-reasoning
[TACL'23] VSR: A probing benchmark for spatial undersranding of vision-language models.
VLMbench
NeurIPS 2022 Paper "VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation"
VL-PLM
Exploiting unlabeled data with vision and language models for object detection, ECCV 2022
synse-zsl
Official PyTorch code for the ICIP 2021 paper 'Syntactically Guided Generative Embeddings For Zero Shot Skeleton Action Recognition'