Autores
Gelbukh Alexander
Título Bi-Bimodal Modality Fusion for Correlation-Controlled Multimodal Sentiment Analysis
Tipo Congreso
Sub-tipo Indefinido
Descripción 23rd ACM International Conference on Multimodal Interaction, ICMI 2021
Resumen Multimodal sentiment analysis aims to extract and integrate semantic information collected from multiple modalities to recognize the expressed emotions and sentiment in multimodal data. This research area's major concern lies in developing an extraordinary fusion scheme that can extract and integrate key information from various modalities. However, previous work is restricted by the lack of leveraging dynamics of independence and correlation between modalities to reach top performance. To mitigate this, we propose the Bi-Bimodal Fusion Network (BBFN), a novel end-to-end network that performs fusion (relevance increment) and separation (difference increment) on pairwise modality representations. The two parts are trained simultaneously such that the combat between them is simulated. The model takes two bimodal pairs as input due to the known information imbalance among modalities. In addition, we leverage a gated control mechanism in the Transformer architecture to further improve the final output. Experimental results on three datasets (CMU-MOSI, CMU-MOSEI, and UR-FUNNY) verifies that our model significantly outperforms the SOTA. The implementation of this work is available at https://github.com/declare-lab/multimodal-deep-learning and https://github.com/declare-lab/BBFN. © 2021 ACM.
Observaciones DOI 10.1145/3462244.3479919
Lugar Virtual, online
País Indefinido
No. de páginas 6-15
Vol. / Cap.
Inicio 2021-10-18
Fin 2021-10-22
ISBN/ISSN 9781450384810