DMER
DMER copied to clipboard
A survey of deep multimodal emotion recognition.
DMER
A survey of deep multimodal emotion recognition.
Performance Comparsion
Summary of latest papers
Update
2022.03.21 Add papers from ACM MM 2021
2022.05.04 Add the pages of performance comparsion and the summary of latest papers.
Structure
- Related Github Repositories
- Datasets
- Related Challenges
- Related Projects
- Related Reviews
- Related Papers
Related Github Repositories
- AWESOME-MER
- AWESOME-MSA
- SER-datasets
- Awesome-Speech-Pretraining
- Reading list for Awesome Sentiment Analysis papers
- Speech-Resources
Datasets
Related Challenges
- Multimodal (Audio, Facial and Gesture) based Emotion Recognition Challenge (MMER) @ FG
- Emotion Recognition in the Wild Challenge (EmotiW) @ ICMI
- Audio/Visual Emotion Challenge (AVEC) @ ACM MM
- One-Minute Gradual-Emotion Behavior Challenge @ IJCNN
- Multimodal Emotion Recognition Challenge (MEC) @ ACII
- Multimodal Pain Recognition (Face and Body) Challenge (EmoPain) @ FG
Related Projects
Related Reviews
- Multimodal Intelligence: Representation Learning, Information Fusion, and Applications--(IEEE Journal of Selected Topics in Signal Processing, 2020)
- A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition--(Information Fusion, 2020)
- Survey on AI-Based Multimodal Methods for Emotion Detection--(High-Performance Modelling and Simulation for Big Data Applications, 2019)
- Multimodal machine learning: A survey and taxonomy--(IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018)
- A review of affective computing: From unimodal analysis to multimodal fusion--(Information Fusion, 2017)
- A survey of multimodal sentiment analysis--(Image and Vision Computing, 2017)
- A Review and Meta-Analysis of Multimodal Affect Detection Systems--(ACM Computing Surveys, 2015)
Video-Audio Method
Context-awarded method
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
CA-1 | EmotiCon | EmotiCon: Context-Aware Multimodal Emotion Recognition using Frege’s Principle | CVPR 2020 | [video] [project] | EMOTIC, [GroupWalk] | Face+Gait+(Depth+Background), Multiplicative fusion, etc |
CA-2 | CAER-Net | Context-Aware Emotion Recognition Networks | ICCV 2019 | [coding][project] | EMOTIC, AffectNet, [CAER-S], AFEW, [CAER] | Face + Context, Adaptive Fusion |
CA-3 | Context-aware affective graph reasoning for emotion recognition | ICME 2019 | ||||
CA-4 | Context Based Emotion Recognition using EMOTIC Dataset | 2019 TPAMI | [coding] | EMOTIC | Face + Context | |
CA-5 | Multimodal Framework for Analyzing the Affect of a Group of People | 2018 TMM | HAPPEI, GAFF | Face+Upper body+Scene, Face-based Group-level Emotion Recognition | ||
CA-6 | Emotion Recognition in Context | CVPR 2017 | [project] | [EMOTIC] | Body feature+Image feature(Context) |
Video-Audio-Text method
Attribute-based
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
AB-1 | MMDRBN | Knowledge-Augmented Multimodal Deep Regression Bayesian Networks for Emotion Video Tagging | 2019 TMM | LIRIS-ACCEDE | Visual + Audio + Attribute | |
AB-2 | Recognizing Induced Emotions of Movie Audiences From Multimodal Information | 2019 TAC | LIRIS-ACCEDE | Visual + Audio + Dialogue + Attribute | ||
AB-3 | Multimodal emotional state recognition using sequence-dependent deep hierarchical features | 2015 Neural Networks | FABO | Face + Upper-body | ||
AB-4 | Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification | 2012 TAC | IEMOCAP | Visual + Audio + Utterance | ||
AB-5 | Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space | 2011 TAC | SAL-DB | Face + Shoulder gesture + Audio |
Aspect-based Network
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
ABN-1 | MIMN | Multi-Interactive Memory Network for Aspect Based Multimodal Sentiment Analysis | AAAI 2019 | [coding] | [Multi-ZOL] | Text+Aspect+Images, Aspect based multimodal sentiment analysis |
ABN-2 | VistaNet | VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis | AAAI 2019 | [coding] | [Yelp-Food-Restaurants] | Visual+Text |
ABN-3 | Cooperative Multimodal Approach to Depression Detection in Twitter | AAAI 2019 | Textual Depression Dataset, Multimodal Depression Dataset | Visual+Text, GRU+VGG-Net+COMMA | ||
ABN-4 | TomBERT | Adapting BERT for Target-Oriented Multimodal Sentiment Classification | IJCAI 2019 | [coding] | Multimodal Twitter datasets | Image+Text, BERT-based |
ABN-5 | Predicting Emotions in User-Generated Videos | AAAI 2014 | [Dataset] | Visual+Audio+Attribute, Video content recognition |