vision-language-model topic
awesome-japanese-llm
日本語LLMまとめ - Overview of Japanese LLMs
llava-docker
Docker image for LLaVA: Large Language and Vision Assistant
RoboFlamingo
Code for RoboFlamingo
InternVL
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源多模态对话模型
Awesome-Multimodal-LLM-Autonomous-Driving
[WACV 2024 Survey Paper] Multimodal Large Language Models for Autonomous Driving
CLIPSelf
[ICLR2024 Spotlight] Code Release of CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction
CG-VLM
This is the official repo for Contrastive Vision-Language Alignment Makes Efficient Instruction Learner.
Txt2Img-MHN
[IEEE TIP 2023] Txt2Img-MHN: Remote Sensing Image Generation from Text Using Modern Hopfield Networks
awesome-knowledge-driven-AD
A curated list of awesome knowledge-driven autonomous driving (continually updated)
LAMDA-PILOT
🎉 PILOT: A Pre-trained Model-Based Continual Learning Toolbox