计算机科学
模态(人机交互)
素描
人工智能
光学(聚焦)
特征(语言学)
情态动词
冗余(工程)
利用
相关性(法律)
特征学习
模式识别(心理学)
特征提取
机器学习
数据挖掘
法学
高分子化学
化学
哲学
算法
物理
光学
操作系统
语言学
计算机安全
政治学
作者
Yue Zhao,Qi Liang,Ruixin Ma,Weizhi Nie,Yuting Su
标识
DOI:10.1016/j.jvcir.2022.103668
摘要
Cross-modal retrieval attracts much research attention due to its wide applications in numerous search systems. Sketch based 3D shape retrieval is a typical challenging cross-modal retrieval task for the huge divergence between sketch modality and 3D shape view modality. Existing approaches project the sketches and shapes into a common space for feature update and data alignment. However, these methods contain several disadvantages: Firstly, the majority approaches ignore the modality-shared information for divergence compensation in descriptor generation process. Secondly, traditional fusion method of multi-view features introduces much redundancy, which decreases the discrimination of shape descriptors. Finally, most approaches only focus on the cross-modal alignment, which omits the modality-specific data relevance. To address these limitations, we propose a Joint Feature Learning Network (JFLN). Firstly, we design a novel modality-shared feature extraction network to exploit both modality-specific characteristics and modality-shared information for descriptor generation. Subsequently, we introduce a hierarchical view attention module to gradually focus on the effective information for multiview feature updating and aggregation. Finally, we propose a novel cross-modal feature learning network, which can simultaneously contribute to modality-specific data distribution and cross-modal data alignment. We conduct exhaustive experiments on three public databases. The experimental results validate the superiority of the proposed method. Full Codes are available at https://github.com/dlmuyy/JFLN.
科研通智能强力驱动
Strongly Powered by AbleSci AI