人工智能
突出
融合
对象(语法)
计算机视觉
计算机科学
模式识别(心理学)
目标检测
传感器融合
哲学
语言学
作者
Yan Liu,Yunzhou Zhang,Zhenyu Wang,Fei Yang,Cao Qin,Feng Qiu,Sonya Coleman,Dermot Kerr
标识
DOI:10.1016/j.imavis.2022.104536
摘要
Salient object detection (SOD) is a challenging and fundamental research in computer vision and image processing. Since the cost of pixel-level annotations is high, scribble annotations are usually used as weak supervisions. However, scribble annotations are too sparse and always located inside the objects with lacking annotations close to the semantic boundaries, which can't make confident predictions. To alleviate these issues, we propose a novel and effective scribble-based weakly supervised approach for SOD, named complementary characteristics fusion network (CCFNet). To be more specific, we design an edge fusion module (EFM) by taking account of local and high-level semantic information to equip our model, which would be beneficial to enhance the power of aggregating edge information. Then to achieve the complementary role of different features, a series of feature correlation modules (FCMs) are employed to strengthen the localization information and details learning. This is based on low-level, high-level global and edge information, which will complement each other to obtain relatively complete salient regions. Alternatively, to encourage the network to learn structural information and further improve the results of saliency maps in foreground and background, we propose a self-supervised salient detection (SSD) loss. Extensive experiments using five benchmark datasets demonstrate that our proposed approach performs favorably against the state-of-the-art weakly supervised algorithms, and even surpasses the performance of those fully supervised.
科研通智能强力驱动
Strongly Powered by AbleSci AI