计算机科学
情绪分析
判别式
人工智能
利用
模态(人机交互)
光学(聚焦)
典型相关
社会化媒体
模式
自然语言处理
模式识别(心理学)
机器学习
社会学
社会科学
万维网
物理
光学
计算机安全
作者
Feiran Huang,Xiaoming Zhang,Zhong-Qiu Zhao,Jie Xu,Zhoujun Li
标识
DOI:10.1016/j.knosys.2019.01.019
摘要
Sentiment analysis of social media data is crucial to understand people’s position, attitude, and opinion toward a certain event, which has many applications such as election prediction and product evaluation. Though great effort has been devoted to the single modality (image or text), less effort is paid to the joint analysis of multimodal data in social media. Most of the existing methods for multimodal sentiment analysis simply combine different data modalities, which results in dissatisfying performance on sentiment classification. In this paper, we propose a novel image–text sentiment analysis model, i.e., Deep Multimodal Attentive Fusion (DMAF), to exploit the discriminative features and the internal correlation between visual and semantic contents with a mixed fusion framework for sentiment analysis. Specifically, to automatically focus on discriminative regions and important words which are most related to the sentiment, two separate unimodal attention models are proposed to learn effective emotion classifiers for visual and textual modality respectively. Then, an intermediate fusion-based multimodal attention model is proposed to exploit the internal correlation between visual and textual features for joint sentiment classification. Finally, a late fusion scheme is applied to combine the three attention models for sentiment prediction. Extensive experiments are conducted to demonstrate the effectiveness of our approach on both weakly labeled and manually labeled datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI