vision-and-language topic
awesome-vln
A curated list of research papers in Vision-Language Navigation (VLN)
AREL
Code for the ACL paper "No Metrics Are Perfect: Adversarial Reward Learning for Visual Storytelling"
pytorch_violet
A PyTorch implementation of VIOLET
VILLA
Research Code for NeurIPS 2020 Spotlight paper "Large-Scale Adversarial Training for Vision-and-Language Representation Learning": UNITER adversarial training part
ALPRO
Align and Prompt: Video-and-Language Pre-training with Entity Prompts
stanford-cs231n-assignments-2020
This repository contains my solutions to the assignments for Stanford's CS231n "Convolutional Neural Networks for Visual Recognition" (Spring 2020).
PhraseCutDataset
Dataset API for "PhraseCut: Language-based Image Segmentation in the Wild"
VAL
Tensorflow Implementation on Paper [CVPR2020]Image Search with Text Feedback by Visiolinguistic Attention Learning
Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
calvin
CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks