Hybrid multi-modal emotion recognition framework based on InceptionV3DenseNet
Abstract
References
Recommendations
Deep learning based multimodal emotion recognition using model-level fusion of audio–visual modalities
AbstractEmotion identification based on multimodal data (e.g., audio, video, text, etc.) is one of the most demanding and important research fields, with various uses. In this context, this research work has conducted a rigorous exploration of ...
Highlights- Deep learning-based feature extractor networks for video and audio data are proposed.
A multimodal emotion recognition method based on multiple fusion of audio-visual modalities
VSIP '23: Proceedings of the 2023 5th International Conference on Video, Signal and Image ProcessingHuman emotions are usually expressed in multiple ways, including speech, facial expressions, body language, etc. However, in multimodal fusion emotion recognition, there are often challenges such as difficulty in data alignment, significant feature ...
Recent developments in openSMILE, the munich open-source multimedia feature extractor
MM '13: Proceedings of the 21st ACM international conference on MultimediaWe present recent developments in the openSMILE feature extraction toolkit. Version 2.0 now unites feature extraction paradigms from speech, music, and general sound events with basic video features for multi-modal processing. Descriptors from audio and ...
Comments
Information & Contributors
Information
Published In
Publisher
Kluwer Academic Publishers
United States
Publication History
Author Tags
Qualifiers
- Research-article
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 0Total Downloads
- Downloads (Last 12 months)0
- Downloads (Last 6 weeks)0