Multi-Modal Attention for Speech Emotion Recognition

FOS: Computer and information sciences Sound (cs.SD) Image and Video Processing (eess.IV) 02 engineering and technology Electrical Engineering and Systems Science - Image and Video Processing Computer Science - Sound Multimedia (cs.MM) Audio and Speech Processing (eess.AS) FOS: Electrical engineering, electronic engineering, information engineering 0202 electrical engineering, electronic engineering, information engineering Computer Science - Multimedia Electrical Engineering and Systems Science - Audio and Speech Processing
DOI: 10.21437/interspeech.2020-1653 Publication Date: 2020-10-27T09:22:11Z
ABSTRACT
Accepted by Interspeech2020<br/>Emotion represents an essential aspect of human speech that is manifested in speech prosody. Speech, visual, and textual cues are complementary in human communication. In this paper, we study a hybrid fusion method, referred to as multi-modal attention network (MMAN) to make use of visual and textual cues in speech emotion recognition. We propose a novel multi-modal attention mechanism, cLSTM-MMA, which facilitates the attention across three modalities and selectively fuse the information. cLSTM-MMA is fused with other uni-modal sub-networks in the late fusion. The experiments show that speech emotion recognition benefits significantly from visual and textual cues, and the proposed cLSTM-MMA alone is as competitive as other fusion methods in terms of accuracy, but with a much more compact network structure. The proposed hybrid network MMAN achieves state-of-the-art performance on IEMOCAP database for emotion recognition.<br/>
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (0)
CITATIONS (45)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....