vision-and-language topic

List vision-and-language repositories

Discrete-Continuous-VLN

76
Stars
7
Forks
Watchers

Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation

TRAR-VQA

63
Stars
18
Forks
Watchers

[ICCV 2021] Official implementation of the paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering"

ZeroVL

44
Stars
5
Forks
Watchers

[ECCV2022] Contrastive Vision-Language Pre-training with Limited Resources

LXMERT-AdvTrain

21
Stars
1
Forks
Watchers

Research Code for NeurIPS 2020 Spotlight paper "Large-Scale Adversarial Training for Vision-and-Language Representation Learning": LXMERT adversarial training part

TVLT

120
Stars
13
Forks
Watchers

PyTorch code for “TVLT: Textless Vision-Language Transformer” (NeurIPS 2022 Oral)

eccv-caption

51
Stars
2
Forks
Watchers

Extended COCO Validation (ECCV) Caption dataset (ECCV 2022)

prismer

1.3k
Stars
75
Forks
Watchers

The implementation of "Prismer: A Vision-Language Model with Multi-Task Experts".

awesome-japanese-llm

962
Stars
29
Forks
Watchers

日本語LLMまとめ - Overview of Japanese LLMs

OFASys

142
Stars
10
Forks
Watchers

OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models

plip

217
Stars
25
Forks
Watchers

Pathology Language and Image Pre-Training (PLIP) is the first vision and language foundation model for Pathology AI (Nature Medicine). PLIP is a large-scale pre-trained model that can be used to extra...