Attention-based multi-level image and text sentiment analysis
Conference: ISCTT 2022 - 7th International Conference on Information Science, Computer Technology and Transportation
05/27/2022 - 05/29/2022 at Xishuangbanna, China
Proceedings: ISCTT 2022
Pages: 6Language: englishTyp: PDF
Authors:
Sun, Xiaomin; Geng, Yushui; Jiang, Wenfeng (School of Computer Science and Technology, Qilu University of Technology (Shandong Academy of Sciences) Jinan, China)
Wu, Qichen (Institute of Computing Technology Beijing Key Laboratory of Mobile Computing and Pervasive Device University of Chinese Academy of Sciences, Beijing, China)
Abstract:
People are keener and keener to express their opinions and share their lives on social media. Analyzing this information with an apparent emotional tendency has broad application prospects. Since the multi-dimensional characteristics of each modality and the interaction between modalities have not been considered in the previous graphic emotion analysis, this paper proposes an emotion analysis model (BRFANet model) based on various neural networks and attention mechanisms. Firstly, ALBERT and BiGRU are used to capture the deep semantic information of the text. Secondly, residual network and Faster RCNN combined with PFN are used to capture the image's global and local features, respectively. Then, to capture the target object that can better reflect emotions, the text features are used to guide attention interaction on local images. Finally, the self-attention network is used to automatically identify the key weights of each mode to complete the fusion. On the Flickr dataset, compared with the traditional multimodal sentiment analysis method, the classification accuracy and F1 value are improved, proving the model's effectiveness.