image-text-retrieval topic
Awesome_Matching_Pretraining_Transfering
The Paper List of Large Multi-Modality Model, Parameter-Efficient Finetuning, Vision-Language Pretraining, Conventional Image-Text Matching for Preliminary Insight.
SGRAF
[AAAI2021] The code of “Similarity Reasoning and Filtration for Image-Text Matching”
tidy
Offline semantic Text-to-Image and Image-to-Image search on Android powered by quantized state-of-the-art vision-language pretrained CLIP model and ONNX Runtime inference engine
mPLUG
mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections. (EMNLP 2022)
RCAR
[TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”
Chinese-CLIP-opencv-onnxrun
使用OpenCV+onnxruntime部署中文clip做以文搜图,给出一句话来描述想要的图片,就能从图库中搜出来符合要求的图片。包含C++和Python两个版本的程序
BagFormer
PyTorch code for BagFormer: Better Cross-Modal Retrieval via bag-wise interaction
Ant-Multi-Modal-Framework
Research Code for Multimodal-Cognition Team in Ant Group
ComCLIP
Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"