Multi-Modal Attention for Speech Emotion Recognition
FOS: Computer and information sciences
Sound (cs.SD)
Image and Video Processing (eess.IV)
02 engineering and technology
Electrical Engineering and Systems Science - Image and Video Processing
Computer Science - Sound
Multimedia (cs.MM)
Audio and Speech Processing (eess.AS)
FOS: Electrical engineering, electronic engineering, information engineering
0202 electrical engineering, electronic engineering, information engineering
Computer Science - Multimedia
Electrical Engineering and Systems Science - Audio and Speech Processing
DOI:
10.21437/interspeech.2020-1653
Publication Date:
2020-10-27T09:22:11Z
AUTHORS (4)
ABSTRACT
Accepted by Interspeech2020<br/>Emotion represents an essential aspect of human speech that is manifested in speech prosody. Speech, visual, and textual cues are complementary in human communication. In this paper, we study a hybrid fusion method, referred to as multi-modal attention network (MMAN) to make use of visual and textual cues in speech emotion recognition. We propose a novel multi-modal attention mechanism, cLSTM-MMA, which facilitates the attention across three modalities and selectively fuse the information. cLSTM-MMA is fused with other uni-modal sub-networks in the late fusion. The experiments show that speech emotion recognition benefits significantly from visual and textual cues, and the proposed cLSTM-MMA alone is as competitive as other fusion methods in terms of accuracy, but with a much more compact network structure. The proposed hybrid network MMAN achieves state-of-the-art performance on IEMOCAP database for emotion recognition.<br/>
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (0)
CITATIONS (45)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....