计算机科学
人工智能
目标检测
分割
模式识别(心理学)
融合机制
背景(考古学)
图像分割
水准点(测量)
比例(比率)
骨干网
特征提取
计算机视觉
上下文模型
边距(机器学习)
机器学习
对象(语法)
融合
哲学
物理
大地测量学
量子力学
脂质双层融合
生物
地理
语言学
计算机网络
古生物学
作者
Yu Liu,Haihang Li,Juan Cheng,Xun Chen
出处
期刊:IEEE Transactions on Circuits and Systems for Video Technology
[Institute of Electrical and Electronics Engineers]
日期:2023-02-16
卷期号:33 (9): 4934-4947
被引量:30
标识
DOI:10.1109/tcsvt.2023.3245883
摘要
The aim of camouflaged object detection (COD) is to find objects that are hidden in their surrounding environment. Due to the factors like low illumination, occlusion, small size and high similarity to the background, COD is recognized to be a very challenging task. In this paper, we propose a general COD framework, termed as MSCAF-Net, focusing on learning multi-scale context-aware features. To achieve this target, we first adopt the improved Pyramid Vision Transformer (PVTv2) model as the backbone to extract global contextual information at multiple scales. An enhanced receptive field (ERF) module is then designed to refine the features at each scale. Further, a cross-scale feature fusion (CSFF) module is introduced to achieve sufficient interaction of multi-scale information, aiming to enrich the scale diversity of extracted features. In addition, inspired the mechanism of the human visual system, a dense interactive decoder (DID) module is devised to output a rough localization map, which is used to modulate the fused features obtained in the CSFF module for more accurate detection. The effectiveness of our MSCAF-Net is validated on four benchmark datasets. The results show that the proposed method significantly outperforms state-of-the-art (SOTA) COD models by a large margin. Besides, we also investigate the potential of our MSCAF-Net on some other vision tasks that are highly related to COD, such as polyp segmentation, COVID-19 lung infection segmentation, transparent object detection and defect detection. Experimental results demonstrate the high versatility of the proposed MSCAF-Net. The source code and results of our method are available at https://github.com/yuliu316316/MSCAF-COD .
科研通智能强力驱动
Strongly Powered by AbleSci AI