计算机科学
互动性
命名实体识别
适应性
人工智能
任务(项目管理)
模式
社会化媒体
构造(python库)
自然语言处理
适应(眼睛)
情报检索
万维网
社会学
经济
物理
管理
程序设计语言
光学
生物
社会科学
生态学
作者
Yu Tian,Xian Sun,Hongfeng Yu,Ya Li,Kun Fu
出处
期刊:Neurocomputing
[Elsevier BV]
日期:2021-01-21
卷期号:439: 12-21
被引量:21
标识
DOI:10.1016/j.neucom.2021.01.060
摘要
Abstract Multimodal Named Entity Recognition task aims to identify named entities in user-generated posts containing both images and texts. Previous multimodal named entity recognition methods greatly benefit from visual features when the text and the image are well aligned, but this is not always the case in social media. On condition that the image is missing or mismatched with the text, these models usually fail to provide excellent performance. Besides, previous models use only single attention to capture the semantic interaction between different modalities, which largely ignore the existence of multiple entity objects in images and texts of the posts. To alleviate these issues, we present a novel model named Hierarchical Self-adaptation Network (HSN) to address these issues. The HSN contains 1) a Cross-modal Interaction Module to promote semantic interaction for the multiple entity objects in different modalities, which is proved to suppress wrong or incomplete attention in multimodal interactivity; 2) a Self-adaptive Multimodal Integration module to handle the problems that the images are missing or mismatched with the texts. Additionally, to evaluate the adaptability of HSN in real-life social media, we construct a Real-world NER dataset consisting of plain text posts and multimodal posts from Twitter. Extensive experiments demonstrate that our model achieves state-of-the-art results on the Real-world multimodal NER dataset and the Twitter multimodal NER dataset.
科研通智能强力驱动
Strongly Powered by AbleSci AI