A survey of deep multimodal emotion recognition.
2022.03.21 Add papers from ACM MM 2021
2022.05.04 Add the pages of performance comparsion and the summary of latest papers.
- Related Github Repositories
- Datasets
- Related Challenges
- Related Projects
- Related Reviews
- Related Papers
- AWESOME-MER
- AWESOME-MSA
- SER-datasets
- Awesome-Speech-Pretraining
- Reading list for Awesome Sentiment Analysis papers
- Speech-Resources
- Multimodal (Audio, Facial and Gesture) based Emotion Recognition Challenge (MMER) @ FG
- Emotion Recognition in the Wild Challenge (EmotiW) @ ICMI
- Audio/Visual Emotion Challenge (AVEC) @ ACM MM
- One-Minute Gradual-Emotion Behavior Challenge @ IJCNN
- Multimodal Emotion Recognition Challenge (MEC) @ ACII
- Multimodal Pain Recognition (Face and Body) Challenge (EmoPain) @ FG
- Multimodal Intelligence: Representation Learning, Information Fusion, and Applications--(IEEE Journal of Selected Topics in Signal Processing, 2020)
- A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition--(Information Fusion, 2020)
- Survey on AI-Based Multimodal Methods for Emotion Detection--(High-Performance Modelling and Simulation for Big Data Applications, 2019)
- Multimodal machine learning: A survey and taxonomy--(IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018)
- A review of affective computing: From unimodal analysis to multimodal fusion--(Information Fusion, 2017)
- A survey of multimodal sentiment analysis--(Image and Vision Computing, 2017)
- A Review and Meta-Analysis of Multimodal Affect Detection Systems--(ACM Computing Surveys, 2015)
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
CA-1 | EmotiCon | EmotiCon: Context-Aware Multimodal Emotion Recognition using Frege’s Principle | CVPR 2020 | [video] [project] | EMOTIC, [GroupWalk] | Face+Gait+(Depth+Background), Multiplicative fusion, etc |
CA-2 | CAER-Net | Context-Aware Emotion Recognition Networks | ICCV 2019 | [coding][project] | EMOTIC, AffectNet, [CAER-S], AFEW, [CAER] | Face + Context, Adaptive Fusion |
CA-3 | Context-aware affective graph reasoning for emotion recognition | ICME 2019 | ||||
CA-4 | Context Based Emotion Recognition using EMOTIC Dataset | 2019 TPAMI | [coding] | EMOTIC | Face + Context | |
CA-5 | Multimodal Framework for Analyzing the Affect of a Group of People | 2018 TMM | HAPPEI, GAFF | Face+Upper body+Scene, Face-based Group-level Emotion Recognition | ||
CA-6 | Emotion Recognition in Context | CVPR 2017 | [project] | [EMOTIC] | Body feature+Image feature(Context) |
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
AB-1 | MMDRBN | Knowledge-Augmented Multimodal Deep Regression Bayesian Networks for Emotion Video Tagging | 2019 TMM | LIRIS-ACCEDE | Visual + Audio + Attribute | |
AB-2 | Recognizing Induced Emotions of Movie Audiences From Multimodal Information | 2019 TAC | LIRIS-ACCEDE | Visual + Audio + Dialogue + Attribute | ||
AB-3 | Multimodal emotional state recognition using sequence-dependent deep hierarchical features | 2015 Neural Networks | FABO | Face + Upper-body | ||
AB-4 | Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification | 2012 TAC | IEMOCAP | Visual + Audio + Utterance | ||
AB-5 | Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space | 2011 TAC | SAL-DB | Face + Shoulder gesture + Audio |
Index | Model | Paper | Year | Project | Dataset | Method |
---|---|---|---|---|---|---|
ABN-1 | MIMN | Multi-Interactive Memory Network for Aspect Based Multimodal Sentiment Analysis | AAAI 2019 | [coding] | [Multi-ZOL] | Text+Aspect+Images, Aspect based multimodal sentiment analysis |
ABN-2 | VistaNet | VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis | AAAI 2019 | [coding] | [Yelp-Food-Restaurants] | Visual+Text |
ABN-3 | Cooperative Multimodal Approach to Depression Detection in Twitter | AAAI 2019 | Textual Depression Dataset, Multimodal Depression Dataset | Visual+Text, GRU+VGG-Net+COMMA | ||
ABN-4 | TomBERT | Adapting BERT for Target-Oriented Multimodal Sentiment Classification | IJCAI 2019 | [coding] | Multimodal Twitter datasets | Image+Text, BERT-based |
ABN-5 | Predicting Emotions in User-Generated Videos | AAAI 2014 | [Dataset] | Visual+Audio+Attribute, Video content recognition |