Exploring Multimodal Sentiment Analysis via CBAM Attention and Double-layer BiLSTM Architecture

03/26/2023
by   Huiru Wang, et al.
0

Because multimodal data contains more modal information, multimodal sentiment analysis has become a recent research hotspot. However, redundant information is easily involved in feature fusion after feature extraction, which has a certain impact on the feature representation after fusion. Therefore, in this papaer, we propose a new multimodal sentiment analysis model. In our model, we use BERT + BiLSTM as new feature extractor to capture the long-distance dependencies in sentences and consider the position information of input sequences to obtain richer text features. To remove redundant information and make the network pay more attention to the correlation between image and text features, CNN and CBAM attention are added after splicing text features and picture features, to improve the feature representation ability. On the MVSA-single dataset and HFM dataset, compared with the baseline model, the ACC of our model is improved by 1.78 3.09 achieves a sound effect, similar to the advanced model.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset