multi-modal-fusion topic
Multi-Modal-Image-Fusion
Training for multi-modal image fusion with PyTorch.
Achelous
Achelous: A Fast Unified Water-surface Panoptic Perception Framework based on Fusion of Monocular Camera and 4D mmWave Radar
the-compiler
Seed, Code, Harvest: Grow Your Own App with Tree of Thoughts!
Husformer
This repository contains the source code for our paper: "Husformer: A Multi-Modal Transformer for Multi-Modal Human State Recognition". For more details, please refer to our paper at https://arxiv.org...
MegaVIT
The open source implementation of the model from "Scaling Vision Transformers to 22 Billion Parameters"
VisEvent_SOT_Benchmark
[IEEE TCYB 2023] The first large-scale tracking dataset by fusing RGB and Event cameras.
ACMVH
Adaptive Confidence Multi-View Hashing
MoE-Mamba
Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Zeta
AM3Net_Multimodal_Data_Fusion
Code for J. Wang, J. Li, Y. Shi, J. Lai and X. Tan, "AM3Net: Adaptive Mutual-learning-based Multimodal Data Fusion Network," in IEEE TCSVT, 2022. We conducted the experiments on the hyperspectral and...
KG-MM-Survey
Knowledge Graphs Meet Multi-Modal Learning: A Comprehensive Survey