Multimodal Sentiment Analysis Based on a Cross-ModalMultihead Attention Mechanism.

Autor: Lujuan Deng, Boyi Liu, Zuhe Li
Předmět:
Zdroj: Computers, Materials & Continua; 2024, Vol. 78 Issue 1, p1157-1170, 14p
Abstrakt: Multimodal sentiment analysis aims to understand people's emotions and opinions from diverse data. Concatenating or multiplying various modalities is a traditional multi-modal sentiment analysis fusion method. This fusion method does not utilize the correlation information between modalities. To solve this problem, this paper proposes amodel based on amulti-head attentionmechanism. First, after preprocessing the original data. Then, the feature representation is converted into a sequence of word vectors and positional encoding is introduced to better understand the semantic and sequential information in the input sequence. Next, the input coding sequence is fed into the transformer model for further processing and learning. At the transformer layer, a cross-modal attention consisting of a pair of multi-head attention modules is employed to reflect the correlation between modalities. Finally, the processed results are input into the feedforward neural network to obtain the emotional output through the classification layer. Through the above processing flow, the model can capture semantic information and contextual relationships and achieve good results in various natural language processing tasks. Our model was tested on the CMU Multimodal Opinion Sentiment and Emotion Intensity (CMU-MOSEI) and Multimodal EmotionLinesDataset (MELD), achieving an accuracy of 82.04% and F1 parameters reached 80.59% on the former dataset. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index